klotz: dimensionality reduction* + pca*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. PCA and t-SNE are popular dimensionality reduction techniques used for data visualization. This tutorial compares PCA and t-SNE, highlighting their strengths and weaknesses, and provides guidance on when to use each method.

    This article from Machine Learning Mastery discusses when to use Principal Component Analysis (PCA) and t-Distributed Stochastic Neighbor Embedding (t-SNE) for dimensionality reduction and data visualization. Here's a summary of the key points:

    * **PCA is a linear dimensionality reduction technique.** It aims to find the directions of greatest variance in the data and project the data onto those directions. It's good for preserving global structure but can distort local relationships. It's computationally efficient.
    * **t-SNE is a non-linear dimensionality reduction technique.** It focuses on preserving the local structure of the data, meaning points that are close together in the high-dimensional space will likely be close together in the low-dimensional space. It excels at revealing clusters but can distort global distances and is computationally expensive.
    * **Key Differences:**
    * **Linearity vs. Non-linearity:** PCA is linear, t-SNE is non-linear.
    * **Global vs. Local Structure:** PCA preserves global structure, t-SNE preserves local structure.
    * **Computational Cost:** PCA is faster, t-SNE is slower.
    * **When to use which:**
    * **PCA:** Use when you need to reduce dimensionality for speed or memory efficiency, and preserving global structure is important. Good for data preprocessing before machine learning algorithms.
    * **t-SNE:** Use when you want to visualize high-dimensional data and reveal clusters, and you're less concerned about preserving global distances. Excellent for exploratory data analysis.
    * **Important Considerations for t-SNE:**
    * **Perplexity:** A key parameter that controls the balance between local and global aspects of the embedding. Experiment with different values.
    * **Randomness:** t-SNE is a stochastic algorithm, so results can vary. Run it multiple times to ensure consistency.
    * **Interpretation:** Distances in the t-SNE plot should not be interpreted as true distances in the original high-dimensional space.



    In essence, the article advises choosing PCA for preserving overall data structure and speed, and t-SNE for revealing clusters and local relationships, understanding its limitations regarding global distance interpretation.
  2. This tutorial demonstrates how to perform document clustering using LLM embeddings with scikit-learn. It covers generating embeddings with Sentence Transformers, reducing dimensionality with PCA, and applying KMeans clustering to group similar documents.
  3. This article details seven advanced feature engineering techniques using LLM embeddings to improve machine learning model performance. It covers techniques like dimensionality reduction, semantic similarity, clustering, and more.

    The article explores how to leverage LLM embeddings for advanced feature engineering in machine learning, going beyond simple similarity searches. It details seven techniques:

    1. **Embedding Arithmetic:** Performing mathematical operations (addition, subtraction) on embeddings to represent concepts like "positive sentiment - negative sentiment = overall sentiment".
    2. **Embedding Clustering:** Using clustering algorithms (like k-means) on embeddings to create categorical features representing groups of similar text.
    3. **Embedding Dimensionality Reduction:** Reducing the dimensionality of embeddings using techniques like PCA or UMAP to create more compact features while preserving important information.
    4. **Embedding as Input to Tree-Based Models:** Directly using embedding vectors as features in tree-based models like Random Forests or Gradient Boosting. The article highlights the importance of careful handling of high-dimensional data.
    5. **Embedding-Weighted Averaging:** Calculating weighted averages of embeddings based on relevance scores (e.g., TF-IDF) to create a single, representative embedding for a document.
    6. **Embedding Difference:** Calculating the difference between embeddings to capture changes or relationships between texts (e.g., before/after edits, question/answer pairs).
    7. **Embedding Concatenation:** Combining multiple embeddings (e.g., title and body of a document) to create a richer feature representation.
  4. PCA (principal component analysis) can be effectively used for outlier detection by transforming data into a space where outliers are more easily identifiable due to the reduction in dimensionality and reshaping of data patterns.
  5. Exploratory data analysis (EDA) is a powerful technique to understand the structure of word embeddings, the basis of large language models. In this article, we'll apply EDA to GloVe word embeddings and find some interesting insights.
  6. This article explains the PCA algorithm and its implementation in Python. It covers key concepts such as Dimensionality Reduction, eigenvectors, and eigenvalues. The tutorial aims to provide a solid understanding of the algorithm's inner workings and its application for dealing with high-dimensional data and the curse of dimensionality.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: dimensionality reduction + pca

About - Propulsed by SemanticScuttle